100 research outputs found
Multiple constraints to compute optical flow
The computation of the optical flow field from an image sequence requires the definition of constraints on the temporal change of image features. In this paper, we consider the implications of using multiple constraints in the computational schema. In the first step, it is shown that differential constraints correspond to an implicit feature tracking. Therefore, the best results (either in terms of measurement accuracy, and speed in the computation) are obtained by selecting and applying the constraints which are best “tuned” to the particular image feature under consideration. Considering also multiple image points not only allows us to obtain a (locally) better estimate of the velocity field, but also to detect erroneous measurements due to discontinuities in the velocity field. Moreover, by hypothesizing a constant acceleration motion model, also the derivatives of the optical flow are computed. Several experiments are presented from real image sequences
Robust multi-modal and multi-unit feature level fusion of face and iris biometrics
Multi-biometrics has recently emerged as a mean of more robust and effcient
personal verification and identification. Exploiting information from multiple
sources at various levels i.e., feature, score, rank or decision, the false acceptance
and rejection rates can be considerably reduced. Among all, feature level fusion
is relatively an understudied problem. This paper addresses the feature level
fusion for multi-modal and multi-unit sources of information. For multi-modal
fusion the face and iris biometric traits are considered, while the multi-unit fusion
is applied to merge the data from the left and right iris images. The proposed
approach computes the SIFT features from both biometric sources, either multi-
modal or multi-unit. For each source, the extracted SIFT features are selected via
spatial sampling. Then these selected features are finally concatenated together
into a single feature super-vector using serial fusion. This concatenated feature
vector is used to perform classification.
Experimental results from face and iris standard biometric databases are
presented. The reported results clearly show the performance improvements in
classification obtained by applying feature level fusion for both multi-modal and
multi-unit biometrics in comparison to uni-modal classification and score level
fusion
Understanding critical factors in gender recognition
Gender classification is a task of paramount importance in face recognition research, and it is potentially useful in a large set of applications. In this paper we investigate the gender classification problem by an extended empirical analysis on the Face Recognition Grand Challenge version 2.0 dataset (FRGC2.0). We propose challenging experimental protocols over the dimensions of FRGC2.0 – i.e., subject, face expression, race, controlled or uncontrolled environment. We evaluate our protocols with respect to several classification algorithms, and processing different types of features, like Gabor and LBP. Our results show that
gender classification is independent from factors like the race of the subject, face expressions, and variations of controlled illumination conditions. We also report that Gabor features seem to be more robust than LBPs in the case of uncontrolled environment
On the use of SIFT features for face authentication
Several pattern recognition and classification techniques
have been applied to the biometrics domain. Among them,
an interesting technique is the Scale Invariant Feature
Transform (SIFT), originally devised for object recognition.
Even if SIFT features have emerged as a very powerful image
descriptors, their employment in face analysis context
has never been systematically investigated.
This paper investigates the application of the SIFT approach
in the context of face authentication. In order to determine
the real potential and applicability of the method,
different matching schemes are proposed and tested using
the BANCA database and protocol, showing promising results
Feature Level Fusion of Face and Fingerprint Biometrics
The aim of this paper is to study the fusion at feature extraction level for
face and fingerprint biometrics. The proposed approach is based on the fusion
of the two traits by extracting independent feature pointsets from the two
modalities, and making the two pointsets compatible for concatenation.
Moreover, to handle the problem of curse of dimensionality, the feature
pointsets are properly reduced in dimension. Different feature reduction
techniques are implemented, prior and after the feature pointsets fusion, and
the results are duly recorded. The fused feature pointset for the database and
the query face and fingerprint images are matched using techniques based on
either the point pattern matching, or the Delaunay triangulation. Comparative
experiments are conducted on chimeric and real databases, to assess the actual
advantage of the fusion performed at the feature extraction level, in
comparison to the matching score level.Comment: 6 pages, 7 figures, conferenc
8. Advanced techniques for face-based biometrics
Face recognition is nowadays one of the most challenging biometric modalities for the identification of individuals. In the last two decades several experimental as well as commercial systems have been developed exploiting different physical properties of the face image. Either being based on processing 2D or 3D information all these methods perform a face classification of the individuals based on some relevant features extracted from the raw image data. The data acquisition, preprocessing and the feature extraction/selection are all topics of the greatest importance to design a good performing recognition system. At the same time, the right choice of the features to be used as the basis for the face representation, which must be based on the uniqueness of such features, as well as most advanced issues such as the incorporation of quality information and the cope for ageing effects, are all of paramount importance. The tutorial will consists of two sessions (half day of total duration) devoted to the description of both the basic and most advanced techniques related to face recognition. The lectures will provide a comprehensive outline of face- based biometrics, its relation to biological systems (the psychophysics of the human visual system), including the existing applications and commercial systems. The lectures will provide an in-depth analysis of the state-of-the-art algorithms for face-image analysis including: face detection and tracking, landmark localization, feature extraction, face representation and classification. The lectures will mainly explore the image processing aspects of the recognition process. As for classification, machine learning algorithms will be also presented, including kernel methods as related to learning and the approximation theory. The most relevant issues and problems will be raised, providing practical solutions and algorithms responding to them. Particular attention will be given to the most advanced and new techniques for face - - representation and classification, as well as the current approaches presented in the literature. Attention will be also given to the performance evaluation of face recognition systems providing some examples and results from recent competitions and public evaluation contests. Finally, the tutorial will present three relevant and novel issues: the use of face image sequences for exploiting the time domain, the extension to 3D face analysis, and the how to cope with ageing and data quality
Visual Surveillance and Biometrics: Practices, Challenges, and Possibilities
Visual surveillance is the latest paradigm for social security through machine intelligence. It includes the use of visual data captured by infrared sensors or visible-light cameras mounted in cars, corridors, traffic signals etc. Visual surveillance facilitates the classification of human behavior, crowd activity, and gesture analysis to achieve application-specific objectivesinfo:eu-repo/semantics/publishedVersio
On the quantitative estimation of short-term aging in human faces
Facial aging has been only partially studied in the past and mostly in a
qualitative way. This paper presents a novel approach to the estimation of facial
aging aimed to the quantitative evaluation of the changes in facial appearance
over time. In particular, the changes both in face shape and texture, due to
short-time aging, are considered. The developed framework exploits the concept
of “distinctiveness” of facial features and the temporal evolution of such measure.
The analysis is performed both at a global and local level to define the features
which are more stable over time.
Several experiments are performed on publicly available databases with image
sequences densely sampled over a time span of several years. The reported results
clearly show the potential of the methodology to a number of applications in
biometric identification from human faces
- …